probability density function
Efficient Nonparametric Smoothness Estimation
Sobolev quantities (norms, inner products, and distances) of probability density functions are important in the theory of nonparametric statistics, but have rarely been used in practice, partly due to a lack of practical estimators. They also include, as special cases, L^2 quantities which are used in many applications. We propose and analyze a family of estimators for Sobolev quantities of unknown probability density functions. We bound the finite-sample bias and variance of our estimators, finding that they are generally minimax rate-optimal. Our estimators are significantly more computationally tractable than previous estimators, and exhibit a statistical/computational trade-off allowing them to adapt to computational constraints. We also draw theoretical connections to recent work on fast two-sample testing and empirically validate our estimators on synthetic data.
A Proof of Theorems
We still need to demonstrate that the properties in P AC-Bayes analysis hold for both the margin operator and the robust margin operator. Then we complete the proof of Lemma 6.1. The proof of Lemma 7.1 and 7.2 is similar. We provide the proof of Lemma 7.2 below. Lemma 7.1 follows the proof of Lemma 7.2 by replacing the robust margin operator by the margin Since the above bound holds for any x in the domain X, we can get the following a.s.: R The second inequality is the tail bound above.
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- Europe > Switzerland > Zürich > Zürich (0.04)
- Asia > Middle East > Jordan (0.04)
- North America > United States > Texas > Travis County > Austin (0.04)
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.04)
- North America > United States > California > San Diego County > San Diego (0.04)
- North America > Canada > Alberta > Census Division No. 15 > Improvement District No. 9 > Banff (0.04)
km(τ) contribute to the node states
WhenT is larger, more recent edges are assignedsmallDAmagnitudes,sothattheessentialsemantic information is preserved. This theorem guarantees that our DA techiniques do not break the original edge time distribution. There are 4,066 drop-out events (= 0.98%). Based on the validation results, using two TGAT layers and two attention heads with dropout rate of 0.1 gives the best performance. For inference, we inductively compute the embeddings for both the unseen and observed nodes at each time point that the graph evolves, or when the node labels are updated.
RPWithPrior: Label Differential Privacy in Regression
With the wide application of machine learning techniques in practice, privacy preservation has gained increasing attention. Protecting user privacy with minimal accuracy loss is a fundamental task in the data analysis and mining community. In this paper, we focus on regression tasks under $ε$-label differential privacy guarantees. Some existing methods for regression with $ε$-label differential privacy, such as the RR-On-Bins mechanism, discretized the output space into finite bins and then applied RR algorithm. To efficiently determine these finite bins, the authors rounded the original responses down to integer values. However, such operations does not align well with real-world scenarios. To overcome these limitations, we model both original and randomized responses as continuous random variables, avoiding discretization entirely. Our novel approach estimates an optimal interval for randomized responses and introduces new algorithms designed for scenarios where a prior is either known or unknown. Additionally, we prove that our algorithm, RPWithPrior, guarantees $ε$-label differential privacy. Numerical results demonstrate that our approach gets better performance compared with the Gaussian, Laplace, Staircase, and RRonBins, Unbiased mechanisms on the Communities and Crime, Criteo Sponsored Search Conversion Log, California Housing datasets.
- North America > United States > California (0.25)
- Asia > China > Hubei Province > Wuhan (0.04)